62 research outputs found

    Multidesigns for a Graph Pair of Order 6

    Get PDF
    Oral presentation abstract

    Decomposing Complete Graphs into a Graph Pair of Order 6

    Get PDF
    Firstly, a graph G consists of a vertex set V (G), and an edge set E (G) of endpoints which relate two vertices with each edge. Also, a decomposition of a graph is a list of subgraphs such that each edge appears in exactly one subgraph in the list. In the field of graph theory, graph decomposition is an active field of research. A graph pair is a pair of graphs on the same vertex set whose union is the complete graph. Abueida and Daven studied decompositions of complete graphs into graph-pairs of order four and five. We are extending their results by investigating which complete graphs decompose into a specific graph pair of order 6

    A Look at Multi-Decompositions of Complete Graphs into Graph Pairs of Order 4

    Get PDF
    Firstly, a graph G consists of a vertex set V (G), and an edge set E (G) of endpoints which relate two vertices with each edge. Also, a decomposition of a graph is a list of subgraphs such that each edge appears in exactly one subgraph in the list. In the field ofgraph theory, graph decomposition is an active field of research. One type of decomposition is graph pairs. A graph pair is a pair of graphs on the same vertex set whose union is the complete graph. Abueida and Daven studieddecompositions of complete graphs into graph-pairs of order four. In their proof, they left a small part to the readers. We will complete this proof

    A Controllable Model of Grounded Response Generation

    Full text link
    Current end-to-end neural conversation models inherently lack the flexibility to impose semantic control in the response generation process, often resulting in uninteresting responses. Attempts to boost informativeness alone come at the expense of factual accuracy, as attested by pretrained language models' propensity to "hallucinate" facts. While this may be mitigated by access to background knowledge, there is scant guarantee of relevance and informativeness in generated responses. We propose a framework that we call controllable grounded response generation (CGRG), in which lexical control phrases are either provided by a user or automatically extracted by a control phrase predictor from dialogue context and grounding knowledge. Quantitative and qualitative results show that, using this framework, a transformer based model with a novel inductive attention mechanism, trained on a conversation-like Reddit dataset, outperforms strong generation baselines.Comment: AAAI 202
    • …
    corecore